2024-12-17 09:09:18,846 INFO mapreduce.MiniHadoopClusterManager: Updated 0 configuration settings from command line. 2024-12-17 09:09:18,897 INFO hdfs.MiniDFSCluster: starting cluster: numNameNodes=1, numDataNodes=1 2024-12-17 09:09:19,280 INFO namenode.NameNode: Formatting using clusterid: testClusterID 2024-12-17 09:09:19,293 INFO namenode.FSEditLog: Edit logging is async:true 2024-12-17 09:09:19,315 INFO namenode.FSNamesystem: KeyProvider: null 2024-12-17 09:09:19,317 INFO namenode.FSNamesystem: fsLock is fair: true 2024-12-17 09:09:19,317 INFO namenode.FSNamesystem: Detailed lock hold time metrics enabled: false 2024-12-17 09:09:19,345 INFO namenode.FSNamesystem: fsOwner = clickhouse (auth:SIMPLE) 2024-12-17 09:09:19,345 INFO namenode.FSNamesystem: supergroup = supergroup 2024-12-17 09:09:19,345 INFO namenode.FSNamesystem: isPermissionEnabled = true 2024-12-17 09:09:19,345 INFO namenode.FSNamesystem: isStoragePolicyEnabled = true 2024-12-17 09:09:19,345 INFO namenode.FSNamesystem: HA Enabled: false 2024-12-17 09:09:19,386 INFO common.Util: dfs.datanode.fileio.profiling.sampling.percentage set to 0. Disabling file IO profiling 2024-12-17 09:09:19,389 INFO Configuration.deprecation: hadoop.configured.node.mapping is deprecated. Instead, use net.topology.configured.node.mapping 2024-12-17 09:09:19,389 INFO blockmanagement.DatanodeManager: dfs.block.invalidate.limit: configured=1000, counted=60, effected=1000 2024-12-17 09:09:19,389 INFO blockmanagement.DatanodeManager: dfs.namenode.datanode.registration.ip-hostname-check=true 2024-12-17 09:09:19,392 INFO blockmanagement.BlockManager: dfs.namenode.startup.delay.block.deletion.sec is set to 000:00:00:00.000 2024-12-17 09:09:19,393 INFO blockmanagement.BlockManager: The block deletion will start around 2024 Dec 17 09:09:19 2024-12-17 09:09:19,394 INFO util.GSet: Computing capacity for map BlocksMap 2024-12-17 09:09:19,394 INFO util.GSet: VM type = 64-bit 2024-12-17 09:09:19,395 INFO util.GSet: 2.0% max memory 7.7 GB = 156.7 MB 2024-12-17 09:09:19,395 INFO util.GSet: capacity = 2^24 = 16777216 entries 2024-12-17 09:09:19,421 INFO blockmanagement.BlockManager: Storage policy satisfier is disabled 2024-12-17 09:09:19,421 INFO blockmanagement.BlockManager: dfs.block.access.token.enable = false 2024-12-17 09:09:19,426 INFO blockmanagement.BlockManagerSafeMode: dfs.namenode.safemode.threshold-pct = 0.999 2024-12-17 09:09:19,426 INFO blockmanagement.BlockManagerSafeMode: dfs.namenode.safemode.min.datanodes = 0 2024-12-17 09:09:19,426 INFO blockmanagement.BlockManagerSafeMode: dfs.namenode.safemode.extension = 0 2024-12-17 09:09:19,426 INFO blockmanagement.BlockManager: defaultReplication = 1 2024-12-17 09:09:19,427 INFO blockmanagement.BlockManager: maxReplication = 512 2024-12-17 09:09:19,427 INFO blockmanagement.BlockManager: minReplication = 1 2024-12-17 09:09:19,427 INFO blockmanagement.BlockManager: maxReplicationStreams = 2 2024-12-17 09:09:19,427 INFO blockmanagement.BlockManager: redundancyRecheckInterval = 3000ms 2024-12-17 09:09:19,427 INFO blockmanagement.BlockManager: encryptDataTransfer = false 2024-12-17 09:09:19,427 INFO blockmanagement.BlockManager: maxNumBlocksToLog = 1000 2024-12-17 09:09:19,443 INFO namenode.FSDirectory: GLOBAL serial map: bits=29 maxEntries=536870911 2024-12-17 09:09:19,443 INFO namenode.FSDirectory: USER serial map: bits=24 maxEntries=16777215 2024-12-17 09:09:19,444 INFO namenode.FSDirectory: GROUP serial map: bits=24 maxEntries=16777215 2024-12-17 09:09:19,444 INFO namenode.FSDirectory: XATTR serial map: bits=24 maxEntries=16777215 2024-12-17 09:09:19,452 INFO util.GSet: Computing capacity for map INodeMap 2024-12-17 09:09:19,452 INFO util.GSet: VM type = 64-bit 2024-12-17 09:09:19,452 INFO util.GSet: 1.0% max memory 7.7 GB = 78.3 MB 2024-12-17 09:09:19,452 INFO util.GSet: capacity = 2^23 = 8388608 entries 2024-12-17 09:09:19,462 INFO namenode.FSDirectory: ACLs enabled? true 2024-12-17 09:09:19,462 INFO namenode.FSDirectory: POSIX ACL inheritance enabled? true 2024-12-17 09:09:19,463 INFO namenode.FSDirectory: XAttrs enabled? true 2024-12-17 09:09:19,463 INFO namenode.NameNode: Caching file names occurring more than 10 times 2024-12-17 09:09:19,466 INFO snapshot.SnapshotManager: Loaded config captureOpenFiles: false, skipCaptureAccessTimeOnlyChange: false, snapshotDiffAllowSnapRootDescendant: true, maxSnapshotLimit: 65536 2024-12-17 09:09:19,467 INFO snapshot.SnapshotManager: SkipList is disabled 2024-12-17 09:09:19,471 INFO util.GSet: Computing capacity for map cachedBlocks 2024-12-17 09:09:19,471 INFO util.GSet: VM type = 64-bit 2024-12-17 09:09:19,471 INFO util.GSet: 0.25% max memory 7.7 GB = 19.6 MB 2024-12-17 09:09:19,471 INFO util.GSet: capacity = 2^21 = 2097152 entries 2024-12-17 09:09:19,479 INFO metrics.TopMetrics: NNTop conf: dfs.namenode.top.window.num.buckets = 10 2024-12-17 09:09:19,479 INFO metrics.TopMetrics: NNTop conf: dfs.namenode.top.num.users = 10 2024-12-17 09:09:19,479 INFO metrics.TopMetrics: NNTop conf: dfs.namenode.top.windows.minutes = 1,5,25 2024-12-17 09:09:19,482 INFO namenode.FSNamesystem: Retry cache on namenode is enabled 2024-12-17 09:09:19,482 INFO namenode.FSNamesystem: Retry cache will use 0.03 of total heap and retry cache entry expiry time is 600000 millis 2024-12-17 09:09:19,483 INFO util.GSet: Computing capacity for map NameNodeRetryCache 2024-12-17 09:09:19,483 INFO util.GSet: VM type = 64-bit 2024-12-17 09:09:19,484 INFO util.GSet: 0.029999999329447746% max memory 7.7 GB = 2.4 MB 2024-12-17 09:09:19,484 INFO util.GSet: capacity = 2^18 = 262144 entries 2024-12-17 09:09:19,502 INFO namenode.FSImage: Allocated new BlockPoolId: BP-1007324782-172.17.0.2-1734390559495 2024-12-17 09:09:19,518 INFO common.Storage: Storage directory /hadoop-3.3.1/target/test/data/dfs/name-0-1 has been successfully formatted. 2024-12-17 09:09:19,520 INFO common.Storage: Storage directory /hadoop-3.3.1/target/test/data/dfs/name-0-2 has been successfully formatted. 2024-12-17 09:09:19,559 INFO namenode.FSImageFormatProtobuf: Saving image file /hadoop-3.3.1/target/test/data/dfs/name-0-1/current/fsimage.ckpt_0000000000000000000 using no compression 2024-12-17 09:09:19,559 INFO namenode.FSImageFormatProtobuf: Saving image file /hadoop-3.3.1/target/test/data/dfs/name-0-2/current/fsimage.ckpt_0000000000000000000 using no compression 2024-12-17 09:09:19,669 INFO namenode.FSImageFormatProtobuf: Image file /hadoop-3.3.1/target/test/data/dfs/name-0-1/current/fsimage.ckpt_0000000000000000000 of size 405 bytes saved in 0 seconds . 2024-12-17 09:09:19,669 INFO namenode.FSImageFormatProtobuf: Image file /hadoop-3.3.1/target/test/data/dfs/name-0-2/current/fsimage.ckpt_0000000000000000000 of size 405 bytes saved in 0 seconds . 2024-12-17 09:09:19,686 INFO namenode.NNStorageRetentionManager: Going to retain 1 images with txid >= 0 2024-12-17 09:09:19,728 INFO namenode.FSNamesystem: Stopping services started for active state 2024-12-17 09:09:19,728 INFO namenode.FSNamesystem: Stopping services started for standby state 2024-12-17 09:09:19,729 INFO namenode.NameNode: createNameNode [] 2024-12-17 09:09:19,784 INFO impl.MetricsConfig: Loaded properties from hadoop-metrics2.properties 2024-12-17 09:09:19,867 INFO impl.MetricsSystemImpl: Scheduled Metric snapshot period at 10 second(s). 2024-12-17 09:09:19,867 INFO impl.MetricsSystemImpl: NameNode metrics system started 2024-12-17 09:09:19,897 INFO namenode.NameNodeUtils: fs.defaultFS is hdfs://127.0.0.1:12222 2024-12-17 09:09:19,897 INFO namenode.NameNode: Clients should use 127.0.0.1:12222 to access this namenode/service. 2024-12-17 09:09:19,942 INFO util.JvmPauseMonitor: Starting JVM pause monitor 2024-12-17 09:09:19,951 INFO hdfs.DFSUtil: Filter initializers set : org.apache.hadoop.http.lib.StaticUserWebFilter,org.apache.hadoop.hdfs.web.AuthFilterInitializer 2024-12-17 09:09:19,955 INFO hdfs.DFSUtil: Starting Web-server for hdfs at: http://localhost:0 2024-12-17 09:09:19,967 INFO util.log: Logging initialized @1563ms to org.eclipse.jetty.util.log.Slf4jLog 2024-12-17 09:09:20,040 INFO server.AuthenticationFilter: Unable to initialize FileSignerSecretProvider, falling back to use random secrets. 2024-12-17 09:09:20,043 INFO http.HttpRequestLog: Http request log for http.requests.namenode is not defined 2024-12-17 09:09:20,050 INFO http.HttpServer2: Added global filter 'safety' (class=org.apache.hadoop.http.HttpServer2$QuotingInputFilter) 2024-12-17 09:09:20,051 INFO http.HttpServer2: Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context hdfs 2024-12-17 09:09:20,052 INFO http.HttpServer2: Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context static 2024-12-17 09:09:20,052 INFO http.HttpServer2: Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context logs 2024-12-17 09:09:20,054 INFO http.HttpServer2: Added filter AuthFilter (class=org.apache.hadoop.hdfs.web.AuthFilter) to context hdfs 2024-12-17 09:09:20,054 INFO http.HttpServer2: Added filter AuthFilter (class=org.apache.hadoop.hdfs.web.AuthFilter) to context static 2024-12-17 09:09:20,054 INFO http.HttpServer2: Added filter AuthFilter (class=org.apache.hadoop.hdfs.web.AuthFilter) to context logs 2024-12-17 09:09:20,084 INFO http.HttpServer2: addJerseyResourcePackage: packageName=org.apache.hadoop.hdfs.server.namenode.web.resources;org.apache.hadoop.hdfs.web.resources, pathSpec=/webhdfs/v1/* 2024-12-17 09:09:20,093 INFO http.HttpServer2: Jetty bound to port 41811 2024-12-17 09:09:20,094 INFO server.Server: jetty-9.4.40.v20210413; built: 2021-04-13T20:42:42.668Z; git: b881a572662e1943a14ae12e7e1207989f218b74; jvm 11.0.24+8-post-Ubuntu-1ubuntu322.04 2024-12-17 09:09:20,118 INFO server.session: DefaultSessionIdManager workerName=node0 2024-12-17 09:09:20,118 INFO server.session: No SessionScavenger set, using defaults 2024-12-17 09:09:20,119 INFO server.session: node0 Scavenging every 600000ms 2024-12-17 09:09:20,131 INFO server.AuthenticationFilter: Unable to initialize FileSignerSecretProvider, falling back to use random secrets. 2024-12-17 09:09:20,133 INFO handler.ContextHandler: Started o.e.j.s.ServletContextHandler@64a1923a{logs,/logs,file:///hadoop-3.3.1/logs,AVAILABLE} 2024-12-17 09:09:20,134 INFO handler.ContextHandler: Started o.e.j.s.ServletContextHandler@50b1f030{static,/static,file:///hadoop-3.3.1/share/hadoop/hdfs/webapps/static/,AVAILABLE} 2024-12-17 09:09:20,195 INFO handler.ContextHandler: Started o.e.j.w.WebAppContext@6075b2d3{hdfs,/,file:///hadoop-3.3.1/share/hadoop/hdfs/webapps/hdfs/,AVAILABLE}{file:/hadoop-3.3.1/share/hadoop/hdfs/webapps/hdfs} 2024-12-17 09:09:20,206 INFO server.AbstractConnector: Started ServerConnector@1d81e101{HTTP/1.1, (http/1.1)}{localhost:41811} 2024-12-17 09:09:20,206 INFO server.Server: Started @1803ms 2024-12-17 09:09:20,211 INFO namenode.FSEditLog: Edit logging is async:true 2024-12-17 09:09:20,224 INFO namenode.FSNamesystem: KeyProvider: null 2024-12-17 09:09:20,224 INFO namenode.FSNamesystem: fsLock is fair: true 2024-12-17 09:09:20,224 INFO namenode.FSNamesystem: Detailed lock hold time metrics enabled: false 2024-12-17 09:09:20,224 INFO namenode.FSNamesystem: fsOwner = clickhouse (auth:SIMPLE) 2024-12-17 09:09:20,225 INFO namenode.FSNamesystem: supergroup = supergroup 2024-12-17 09:09:20,225 INFO namenode.FSNamesystem: isPermissionEnabled = true 2024-12-17 09:09:20,225 INFO namenode.FSNamesystem: isStoragePolicyEnabled = true 2024-12-17 09:09:20,225 INFO namenode.FSNamesystem: HA Enabled: false 2024-12-17 09:09:20,225 INFO common.Util: dfs.datanode.fileio.profiling.sampling.percentage set to 0. Disabling file IO profiling 2024-12-17 09:09:20,225 INFO blockmanagement.DatanodeManager: dfs.block.invalidate.limit: configured=1000, counted=60, effected=1000 2024-12-17 09:09:20,226 INFO blockmanagement.DatanodeManager: dfs.namenode.datanode.registration.ip-hostname-check=true 2024-12-17 09:09:20,226 INFO blockmanagement.BlockManager: dfs.namenode.startup.delay.block.deletion.sec is set to 000:00:00:00.000 2024-12-17 09:09:20,226 INFO blockmanagement.BlockManager: The block deletion will start around 2024 Dec 17 09:09:20 2024-12-17 09:09:20,226 INFO util.GSet: Computing capacity for map BlocksMap 2024-12-17 09:09:20,226 INFO util.GSet: VM type = 64-bit 2024-12-17 09:09:20,226 INFO util.GSet: 2.0% max memory 7.7 GB = 156.7 MB 2024-12-17 09:09:20,226 INFO util.GSet: capacity = 2^24 = 16777216 entries 2024-12-17 09:09:20,231 INFO blockmanagement.BlockManager: Storage policy satisfier is disabled 2024-12-17 09:09:20,231 INFO blockmanagement.BlockManager: dfs.block.access.token.enable = false 2024-12-17 09:09:20,231 INFO blockmanagement.BlockManagerSafeMode: dfs.namenode.safemode.threshold-pct = 0.999 2024-12-17 09:09:20,231 INFO blockmanagement.BlockManagerSafeMode: dfs.namenode.safemode.min.datanodes = 0 2024-12-17 09:09:20,231 INFO blockmanagement.BlockManagerSafeMode: dfs.namenode.safemode.extension = 0 2024-12-17 09:09:20,231 INFO blockmanagement.BlockManager: defaultReplication = 1 2024-12-17 09:09:20,231 INFO blockmanagement.BlockManager: maxReplication = 512 2024-12-17 09:09:20,231 INFO blockmanagement.BlockManager: minReplication = 1 2024-12-17 09:09:20,231 INFO blockmanagement.BlockManager: maxReplicationStreams = 2 2024-12-17 09:09:20,231 INFO blockmanagement.BlockManager: redundancyRecheckInterval = 3000ms 2024-12-17 09:09:20,231 INFO blockmanagement.BlockManager: encryptDataTransfer = false 2024-12-17 09:09:20,231 INFO blockmanagement.BlockManager: maxNumBlocksToLog = 1000 2024-12-17 09:09:20,231 INFO util.GSet: Computing capacity for map INodeMap 2024-12-17 09:09:20,231 INFO util.GSet: VM type = 64-bit 2024-12-17 09:09:20,232 INFO util.GSet: 1.0% max memory 7.7 GB = 78.3 MB 2024-12-17 09:09:20,232 INFO util.GSet: capacity = 2^23 = 8388608 entries 2024-12-17 09:09:20,243 INFO namenode.FSDirectory: ACLs enabled? true 2024-12-17 09:09:20,243 INFO namenode.FSDirectory: POSIX ACL inheritance enabled? true 2024-12-17 09:09:20,243 INFO namenode.FSDirectory: XAttrs enabled? true 2024-12-17 09:09:20,244 INFO namenode.NameNode: Caching file names occurring more than 10 times 2024-12-17 09:09:20,244 INFO snapshot.SnapshotManager: Loaded config captureOpenFiles: false, skipCaptureAccessTimeOnlyChange: false, snapshotDiffAllowSnapRootDescendant: true, maxSnapshotLimit: 65536 2024-12-17 09:09:20,244 INFO snapshot.SnapshotManager: SkipList is disabled 2024-12-17 09:09:20,244 INFO util.GSet: Computing capacity for map cachedBlocks 2024-12-17 09:09:20,244 INFO util.GSet: VM type = 64-bit 2024-12-17 09:09:20,244 INFO util.GSet: 0.25% max memory 7.7 GB = 19.6 MB 2024-12-17 09:09:20,244 INFO util.GSet: capacity = 2^21 = 2097152 entries 2024-12-17 09:09:20,247 INFO metrics.TopMetrics: NNTop conf: dfs.namenode.top.window.num.buckets = 10 2024-12-17 09:09:20,247 INFO metrics.TopMetrics: NNTop conf: dfs.namenode.top.num.users = 10 2024-12-17 09:09:20,247 INFO metrics.TopMetrics: NNTop conf: dfs.namenode.top.windows.minutes = 1,5,25 2024-12-17 09:09:20,247 INFO namenode.FSNamesystem: Retry cache on namenode is enabled 2024-12-17 09:09:20,247 INFO namenode.FSNamesystem: Retry cache will use 0.03 of total heap and retry cache entry expiry time is 600000 millis 2024-12-17 09:09:20,247 INFO util.GSet: Computing capacity for map NameNodeRetryCache 2024-12-17 09:09:20,247 INFO util.GSet: VM type = 64-bit 2024-12-17 09:09:20,247 INFO util.GSet: 0.029999999329447746% max memory 7.7 GB = 2.4 MB 2024-12-17 09:09:20,248 INFO util.GSet: capacity = 2^18 = 262144 entries 2024-12-17 09:09:20,256 INFO common.Storage: Lock on /hadoop-3.3.1/target/test/data/dfs/name-0-1/in_use.lock acquired by nodename 484@b6872255a0bb 2024-12-17 09:09:20,258 INFO common.Storage: Lock on /hadoop-3.3.1/target/test/data/dfs/name-0-2/in_use.lock acquired by nodename 484@b6872255a0bb 2024-12-17 09:09:20,260 INFO namenode.FileJournalManager: Recovering unfinalized segments in /hadoop-3.3.1/target/test/data/dfs/name-0-1/current 2024-12-17 09:09:20,260 INFO namenode.FileJournalManager: Recovering unfinalized segments in /hadoop-3.3.1/target/test/data/dfs/name-0-2/current 2024-12-17 09:09:20,260 INFO namenode.FSImage: No edit log streams selected. 2024-12-17 09:09:20,260 INFO namenode.FSImage: Planning to load image: FSImageFile(file=/hadoop-3.3.1/target/test/data/dfs/name-0-1/current/fsimage_0000000000000000000, cpktTxId=0000000000000000000) 2024-12-17 09:09:20,276 INFO namenode.FSImageFormatPBINode: Loading 1 INodes. 2024-12-17 09:09:20,276 INFO namenode.FSImageFormatPBINode: Successfully loaded 1 inodes 2024-12-17 09:09:20,279 INFO namenode.FSImageFormatPBINode: Completed update blocks map and name cache, total waiting duration 0ms. 2024-12-17 09:09:20,281 INFO namenode.FSImageFormatProtobuf: Loaded FSImage in 0 seconds. 2024-12-17 09:09:20,281 INFO namenode.FSImage: Loaded image for txid 0 from /hadoop-3.3.1/target/test/data/dfs/name-0-1/current/fsimage_0000000000000000000 2024-12-17 09:09:20,284 INFO namenode.FSNamesystem: Need to save fs image? false (staleImage=false, haEnabled=false, isRollingUpgrade=false) 2024-12-17 09:09:20,284 INFO namenode.FSEditLog: Starting log segment at 1 2024-12-17 09:09:20,298 INFO namenode.NameCache: initialized with 0 entries 0 lookups 2024-12-17 09:09:20,298 INFO namenode.FSNamesystem: Finished loading FSImage in 49 msecs 2024-12-17 09:09:20,420 INFO namenode.NameNode: RPC server is binding to localhost:12222 2024-12-17 09:09:20,420 INFO namenode.NameNode: Enable NameNode state context:false 2024-12-17 09:09:20,427 INFO ipc.CallQueueManager: Using callQueue: class java.util.concurrent.LinkedBlockingQueue, queueCapacity: 1000, scheduler: class org.apache.hadoop.ipc.DefaultRpcScheduler, ipcBackoff: false. 2024-12-17 09:09:20,435 INFO ipc.Server: Starting Socket Reader #1 for port 12222 2024-12-17 09:09:20,600 INFO namenode.FSNamesystem: Registered FSNamesystemState, ReplicatedBlocksState and ECBlockGroupsState MBeans. 2024-12-17 09:09:20,609 INFO namenode.LeaseManager: Number of blocks under construction: 0 2024-12-17 09:09:20,615 INFO blockmanagement.DatanodeAdminDefaultMonitor: Initialized the Default Decommission and Maintenance monitor 2024-12-17 09:09:20,617 INFO blockmanagement.BlockManager: initializing replication queues 2024-12-17 09:09:20,618 INFO hdfs.StateChange: STATE* Leaving safe mode after 0 secs 2024-12-17 09:09:20,618 INFO hdfs.StateChange: STATE* Network topology has 0 racks and 0 datanodes 2024-12-17 09:09:20,618 INFO hdfs.StateChange: STATE* UnderReplicatedBlocks has 0 blocks 2024-12-17 09:09:20,625 INFO blockmanagement.BlockManager: Total number of blocks = 0 2024-12-17 09:09:20,626 INFO blockmanagement.BlockManager: Number of invalid blocks = 0 2024-12-17 09:09:20,626 INFO blockmanagement.BlockManager: Number of under-replicated blocks = 0 2024-12-17 09:09:20,626 INFO blockmanagement.BlockManager: Number of over-replicated blocks = 0 2024-12-17 09:09:20,626 INFO blockmanagement.BlockManager: Number of blocks being written = 0 2024-12-17 09:09:20,626 INFO hdfs.StateChange: STATE* Replication Queue initialization scan for invalid, over- and under-replicated blocks completed in 8 msec 2024-12-17 09:09:20,646 INFO ipc.Server: IPC Server Responder: starting 2024-12-17 09:09:20,647 INFO ipc.Server: IPC Server listener on 12222: starting 2024-12-17 09:09:20,649 INFO namenode.NameNode: NameNode RPC up at: localhost/127.0.0.1:12222 2024-12-17 09:09:20,651 INFO namenode.FSNamesystem: Starting services required for active state 2024-12-17 09:09:20,651 INFO namenode.FSDirectory: Initializing quota with 12 thread(s) 2024-12-17 09:09:20,655 INFO namenode.FSDirectory: Quota initialization completed in 3 milliseconds name space=1 storage space=0 storage types=RAM_DISK=0, SSD=0, DISK=0, ARCHIVE=0, PROVIDED=0 2024-12-17 09:09:20,659 INFO blockmanagement.CacheReplicationMonitor: Starting CacheReplicationMonitor with interval 30000 milliseconds 2024-12-17 09:09:20,665 INFO hdfs.MiniDFSCluster: Starting DataNode 0 with dfs.datanode.data.dir: [DISK]file:/hadoop-3.3.1/target/test/data/dfs/data/data1,[DISK]file:/hadoop-3.3.1/target/test/data/dfs/data/data2 2024-12-17 09:09:20,706 INFO checker.ThrottledAsyncChecker: Scheduling a check for [DISK]file:/hadoop-3.3.1/target/test/data/dfs/data/data1 2024-12-17 09:09:20,714 INFO checker.ThrottledAsyncChecker: Scheduling a check for [DISK]file:/hadoop-3.3.1/target/test/data/dfs/data/data2 2024-12-17 09:09:20,737 INFO impl.MetricsSystemImpl: DataNode metrics system started (again) 2024-12-17 09:09:20,741 INFO common.Util: dfs.datanode.fileio.profiling.sampling.percentage set to 0. Disabling file IO profiling 2024-12-17 09:09:20,744 INFO datanode.BlockScanner: Initialized block scanner with targetBytesPerSec 1048576 2024-12-17 09:09:20,747 INFO datanode.DataNode: Configured hostname is 127.0.0.1 2024-12-17 09:09:20,748 INFO common.Util: dfs.datanode.fileio.profiling.sampling.percentage set to 0. Disabling file IO profiling 2024-12-17 09:09:20,750 INFO datanode.DataNode: Starting DataNode with maxLockedMemory = 0 2024-12-17 09:09:20,754 INFO datanode.DataNode: Opened streaming server at /127.0.0.1:33271 2024-12-17 09:09:20,756 INFO datanode.DataNode: Balancing bandwidth is 104857600 bytes/s 2024-12-17 09:09:20,756 INFO datanode.DataNode: Number threads for balancing is 100 2024-12-17 09:09:20,780 INFO server.AuthenticationFilter: Unable to initialize FileSignerSecretProvider, falling back to use random secrets. 2024-12-17 09:09:20,781 INFO http.HttpRequestLog: Http request log for http.requests.datanode is not defined 2024-12-17 09:09:20,783 INFO http.HttpServer2: Added global filter 'safety' (class=org.apache.hadoop.http.HttpServer2$QuotingInputFilter) 2024-12-17 09:09:20,784 INFO http.HttpServer2: Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context datanode 2024-12-17 09:09:20,784 INFO http.HttpServer2: Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context logs 2024-12-17 09:09:20,784 INFO http.HttpServer2: Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context static 2024-12-17 09:09:20,787 INFO http.HttpServer2: Jetty bound to port 36427 2024-12-17 09:09:20,787 INFO server.Server: jetty-9.4.40.v20210413; built: 2021-04-13T20:42:42.668Z; git: b881a572662e1943a14ae12e7e1207989f218b74; jvm 11.0.24+8-post-Ubuntu-1ubuntu322.04 2024-12-17 09:09:20,788 INFO server.session: DefaultSessionIdManager workerName=node0 2024-12-17 09:09:20,788 INFO server.session: No SessionScavenger set, using defaults 2024-12-17 09:09:20,789 INFO server.session: node0 Scavenging every 660000ms 2024-12-17 09:09:20,789 INFO handler.ContextHandler: Started o.e.j.s.ServletContextHandler@54db056b{logs,/logs,file:///hadoop-3.3.1/logs,AVAILABLE} 2024-12-17 09:09:20,790 INFO handler.ContextHandler: Started o.e.j.s.ServletContextHandler@5981f4a6{static,/static,file:///hadoop-3.3.1/share/hadoop/hdfs/webapps/static/,AVAILABLE} 2024-12-17 09:09:20,798 INFO handler.ContextHandler: Started o.e.j.w.WebAppContext@16a9eb2e{datanode,/,file:///hadoop-3.3.1/share/hadoop/hdfs/webapps/datanode/,AVAILABLE}{file:/hadoop-3.3.1/share/hadoop/hdfs/webapps/datanode} 2024-12-17 09:09:20,799 INFO server.AbstractConnector: Started ServerConnector@f5b6e78{HTTP/1.1, (http/1.1)}{localhost:36427} 2024-12-17 09:09:20,799 INFO server.Server: Started @2395ms 2024-12-17 09:09:20,891 WARN web.DatanodeHttpServer: Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-17 09:09:20,954 INFO web.DatanodeHttpServer: Listening HTTP traffic on /127.0.0.1:41033 2024-12-17 09:09:20,955 INFO util.JvmPauseMonitor: Starting JVM pause monitor 2024-12-17 09:09:20,956 INFO datanode.DataNode: dnUserName = clickhouse 2024-12-17 09:09:20,956 INFO datanode.DataNode: supergroup = supergroup 2024-12-17 09:09:20,967 INFO ipc.CallQueueManager: Using callQueue: class java.util.concurrent.LinkedBlockingQueue, queueCapacity: 1000, scheduler: class org.apache.hadoop.ipc.DefaultRpcScheduler, ipcBackoff: false. 2024-12-17 09:09:20,968 INFO ipc.Server: Starting Socket Reader #1 for port 0 2024-12-17 09:09:20,973 INFO datanode.DataNode: Opened IPC server at /127.0.0.1:43389 2024-12-17 09:09:20,996 INFO datanode.DataNode: Refresh request received for nameservices: null 2024-12-17 09:09:20,998 INFO datanode.DataNode: Starting BPOfferServices for nameservices: 2024-12-17 09:09:21,005 INFO datanode.DataNode: Block pool (Datanode Uuid unassigned) service to localhost/127.0.0.1:12222 starting to offer service 2024-12-17 09:09:21,012 INFO ipc.Server: IPC Server Responder: starting 2024-12-17 09:09:21,012 INFO ipc.Server: IPC Server listener on 0: starting 2024-12-17 09:09:21,156 INFO datanode.DataNode: Acknowledging ACTIVE Namenode during handshakeBlock pool (Datanode Uuid unassigned) service to localhost/127.0.0.1:12222 2024-12-17 09:09:21,159 INFO common.Storage: Using 2 threads to upgrade data directories (dfs.datanode.parallel.volumes.load.threads.num=2, dataDirs=2) 2024-12-17 09:09:21,162 INFO common.Storage: Lock on /hadoop-3.3.1/target/test/data/dfs/data/data1/in_use.lock acquired by nodename 484@b6872255a0bb 2024-12-17 09:09:21,162 INFO common.Storage: Storage directory with location [DISK]file:/hadoop-3.3.1/target/test/data/dfs/data/data1 is not formatted for namespace 1931591070. Formatting... 2024-12-17 09:09:21,163 INFO common.Storage: Generated new storageID DS-ff9e683c-0e29-43e9-8929-08b6041aaf16 for directory /hadoop-3.3.1/target/test/data/dfs/data/data1 2024-12-17 09:09:21,168 INFO common.Storage: Lock on /hadoop-3.3.1/target/test/data/dfs/data/data2/in_use.lock acquired by nodename 484@b6872255a0bb 2024-12-17 09:09:21,168 INFO common.Storage: Storage directory with location [DISK]file:/hadoop-3.3.1/target/test/data/dfs/data/data2 is not formatted for namespace 1931591070. Formatting... 2024-12-17 09:09:21,168 INFO common.Storage: Generated new storageID DS-cb7c3e35-f4f6-4e0e-80d1-8cca01bf4884 for directory /hadoop-3.3.1/target/test/data/dfs/data/data2 2024-12-17 09:09:21,189 INFO common.Storage: Analyzing storage directories for bpid BP-1007324782-172.17.0.2-1734390559495 2024-12-17 09:09:21,190 INFO common.Storage: Locking is disabled for /hadoop-3.3.1/target/test/data/dfs/data/data1/current/BP-1007324782-172.17.0.2-1734390559495 2024-12-17 09:09:21,190 INFO common.Storage: Block pool storage directory for location [DISK]file:/hadoop-3.3.1/target/test/data/dfs/data/data1 and block pool id BP-1007324782-172.17.0.2-1734390559495 is not formatted. Formatting ... 2024-12-17 09:09:21,190 INFO common.Storage: Formatting block pool BP-1007324782-172.17.0.2-1734390559495 directory /hadoop-3.3.1/target/test/data/dfs/data/data1/current/BP-1007324782-172.17.0.2-1734390559495/current 2024-12-17 09:09:21,208 INFO common.Storage: Analyzing storage directories for bpid BP-1007324782-172.17.0.2-1734390559495 2024-12-17 09:09:21,208 INFO common.Storage: Locking is disabled for /hadoop-3.3.1/target/test/data/dfs/data/data2/current/BP-1007324782-172.17.0.2-1734390559495 2024-12-17 09:09:21,208 INFO common.Storage: Block pool storage directory for location [DISK]file:/hadoop-3.3.1/target/test/data/dfs/data/data2 and block pool id BP-1007324782-172.17.0.2-1734390559495 is not formatted. Formatting ... 2024-12-17 09:09:21,208 INFO common.Storage: Formatting block pool BP-1007324782-172.17.0.2-1734390559495 directory /hadoop-3.3.1/target/test/data/dfs/data/data2/current/BP-1007324782-172.17.0.2-1734390559495/current 2024-12-17 09:09:21,211 INFO datanode.DataNode: Setting up storage: nsid=1931591070;bpid=BP-1007324782-172.17.0.2-1734390559495;lv=-57;nsInfo=lv=-66;cid=testClusterID;nsid=1931591070;c=1734390559495;bpid=BP-1007324782-172.17.0.2-1734390559495;dnuuid=null 2024-12-17 09:09:21,213 INFO datanode.DataNode: Generated and persisted new Datanode UUID a208327e-a2f5-4ce9-88fe-d17d4b676b80 2024-12-17 09:09:21,223 INFO impl.FsDatasetImpl: The datanode lock is a read write lock 2024-12-17 09:09:21,283 INFO impl.FsDatasetImpl: Added new volume: DS-ff9e683c-0e29-43e9-8929-08b6041aaf16 2024-12-17 09:09:21,283 INFO impl.FsDatasetImpl: Added volume - [DISK]file:/hadoop-3.3.1/target/test/data/dfs/data/data1, StorageType: DISK 2024-12-17 09:09:21,286 INFO impl.FsDatasetImpl: Added new volume: DS-cb7c3e35-f4f6-4e0e-80d1-8cca01bf4884 2024-12-17 09:09:21,287 INFO impl.FsDatasetImpl: Added volume - [DISK]file:/hadoop-3.3.1/target/test/data/dfs/data/data2, StorageType: DISK 2024-12-17 09:09:21,291 INFO impl.MemoryMappableBlockLoader: Initializing cache loader: MemoryMappableBlockLoader. 2024-12-17 09:09:21,293 INFO impl.FsDatasetImpl: Registered FSDatasetState MBean 2024-12-17 09:09:21,298 INFO impl.FsDatasetImpl: Adding block pool BP-1007324782-172.17.0.2-1734390559495 2024-12-17 09:09:21,299 INFO impl.FsDatasetImpl: Scanning block pool BP-1007324782-172.17.0.2-1734390559495 on volume /hadoop-3.3.1/target/test/data/dfs/data/data1... 2024-12-17 09:09:21,299 INFO impl.FsDatasetImpl: Scanning block pool BP-1007324782-172.17.0.2-1734390559495 on volume /hadoop-3.3.1/target/test/data/dfs/data/data2... 2024-12-17 09:09:21,311 INFO hdfs.MiniDFSCluster: dnInfo.length != numDataNodes 2024-12-17 09:09:21,312 INFO hdfs.MiniDFSCluster: Waiting for cluster to become active 2024-12-17 09:09:21,339 INFO impl.FsDatasetImpl: Time taken to scan block pool BP-1007324782-172.17.0.2-1734390559495 on /hadoop-3.3.1/target/test/data/dfs/data/data2: 40ms 2024-12-17 09:09:21,339 INFO impl.FsDatasetImpl: Time taken to scan block pool BP-1007324782-172.17.0.2-1734390559495 on /hadoop-3.3.1/target/test/data/dfs/data/data1: 40ms 2024-12-17 09:09:21,339 INFO impl.FsDatasetImpl: Total time to scan all replicas for block pool BP-1007324782-172.17.0.2-1734390559495: 41ms 2024-12-17 09:09:21,340 INFO impl.FsDatasetImpl: Adding replicas to map for block pool BP-1007324782-172.17.0.2-1734390559495 on volume /hadoop-3.3.1/target/test/data/dfs/data/data1... 2024-12-17 09:09:21,340 INFO impl.FsDatasetImpl: Adding replicas to map for block pool BP-1007324782-172.17.0.2-1734390559495 on volume /hadoop-3.3.1/target/test/data/dfs/data/data2... 2024-12-17 09:09:21,340 INFO impl.BlockPoolSlice: Replica Cache file: /hadoop-3.3.1/target/test/data/dfs/data/data1/current/BP-1007324782-172.17.0.2-1734390559495/current/replicas doesn't exist 2024-12-17 09:09:21,340 INFO impl.BlockPoolSlice: Replica Cache file: /hadoop-3.3.1/target/test/data/dfs/data/data2/current/BP-1007324782-172.17.0.2-1734390559495/current/replicas doesn't exist 2024-12-17 09:09:21,342 INFO impl.FsDatasetImpl: Time to add replicas to map for block pool BP-1007324782-172.17.0.2-1734390559495 on volume /hadoop-3.3.1/target/test/data/dfs/data/data1: 1ms 2024-12-17 09:09:21,342 INFO impl.FsDatasetImpl: Time to add replicas to map for block pool BP-1007324782-172.17.0.2-1734390559495 on volume /hadoop-3.3.1/target/test/data/dfs/data/data2: 1ms 2024-12-17 09:09:21,342 INFO impl.FsDatasetImpl: Total time to add all replicas to map for block pool BP-1007324782-172.17.0.2-1734390559495: 3ms 2024-12-17 09:09:21,342 INFO checker.ThrottledAsyncChecker: Scheduling a check for /hadoop-3.3.1/target/test/data/dfs/data/data1 2024-12-17 09:09:21,351 INFO checker.DatasetVolumeChecker: Scheduled health check for volume /hadoop-3.3.1/target/test/data/dfs/data/data1 2024-12-17 09:09:21,352 INFO checker.ThrottledAsyncChecker: Scheduling a check for /hadoop-3.3.1/target/test/data/dfs/data/data2 2024-12-17 09:09:21,353 INFO checker.DatasetVolumeChecker: Scheduled health check for volume /hadoop-3.3.1/target/test/data/dfs/data/data2 2024-12-17 09:09:21,355 INFO datanode.VolumeScanner: Now scanning bpid BP-1007324782-172.17.0.2-1734390559495 on volume /hadoop-3.3.1/target/test/data/dfs/data/data2 2024-12-17 09:09:21,355 INFO datanode.VolumeScanner: Now scanning bpid BP-1007324782-172.17.0.2-1734390559495 on volume /hadoop-3.3.1/target/test/data/dfs/data/data1 2024-12-17 09:09:21,356 INFO datanode.VolumeScanner: VolumeScanner(/hadoop-3.3.1/target/test/data/dfs/data/data1, DS-ff9e683c-0e29-43e9-8929-08b6041aaf16): finished scanning block pool BP-1007324782-172.17.0.2-1734390559495 2024-12-17 09:09:21,356 INFO datanode.VolumeScanner: VolumeScanner(/hadoop-3.3.1/target/test/data/dfs/data/data2, DS-cb7c3e35-f4f6-4e0e-80d1-8cca01bf4884): finished scanning block pool BP-1007324782-172.17.0.2-1734390559495 2024-12-17 09:09:21,357 WARN datanode.DirectoryScanner: dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-17 09:09:21,357 INFO datanode.DirectoryScanner: Periodic Directory Tree Verification scan starting in 5973628ms with interval of 21600000ms and throttle limit of -1ms/s 2024-12-17 09:09:21,363 INFO datanode.DataNode: Block pool BP-1007324782-172.17.0.2-1734390559495 (Datanode Uuid a208327e-a2f5-4ce9-88fe-d17d4b676b80) service to localhost/127.0.0.1:12222 beginning handshake with NN 2024-12-17 09:09:21,367 INFO datanode.VolumeScanner: VolumeScanner(/hadoop-3.3.1/target/test/data/dfs/data/data1, DS-ff9e683c-0e29-43e9-8929-08b6041aaf16): no suitable block pools found to scan. Waiting 1814399988 ms. 2024-12-17 09:09:21,367 INFO datanode.VolumeScanner: VolumeScanner(/hadoop-3.3.1/target/test/data/dfs/data/data2, DS-cb7c3e35-f4f6-4e0e-80d1-8cca01bf4884): no suitable block pools found to scan. Waiting 1814399988 ms. 2024-12-17 09:09:21,380 INFO hdfs.StateChange: BLOCK* registerDatanode: from DatanodeRegistration(127.0.0.1:33271, datanodeUuid=a208327e-a2f5-4ce9-88fe-d17d4b676b80, infoPort=41033, infoSecurePort=0, ipcPort=43389, storageInfo=lv=-57;cid=testClusterID;nsid=1931591070;c=1734390559495) storage a208327e-a2f5-4ce9-88fe-d17d4b676b80 2024-12-17 09:09:21,381 INFO net.NetworkTopology: Adding a new node: /default-rack/127.0.0.1:33271 2024-12-17 09:09:21,381 INFO blockmanagement.BlockReportLeaseManager: Registered DN a208327e-a2f5-4ce9-88fe-d17d4b676b80 (127.0.0.1:33271). 2024-12-17 09:09:21,386 INFO datanode.DataNode: Block pool BP-1007324782-172.17.0.2-1734390559495 (Datanode Uuid a208327e-a2f5-4ce9-88fe-d17d4b676b80) service to localhost/127.0.0.1:12222 successfully registered with NN 2024-12-17 09:09:21,386 INFO datanode.DataNode: For namenode localhost/127.0.0.1:12222 using BLOCKREPORT_INTERVAL of 21600000msecs CACHEREPORT_INTERVAL of 10000msecs Initial delay: 0msecs; heartBeatInterval=3000 2024-12-17 09:09:21,403 INFO blockmanagement.DatanodeDescriptor: Adding new storage ID DS-ff9e683c-0e29-43e9-8929-08b6041aaf16 for DN 127.0.0.1:33271 2024-12-17 09:09:21,403 INFO blockmanagement.DatanodeDescriptor: Adding new storage ID DS-cb7c3e35-f4f6-4e0e-80d1-8cca01bf4884 for DN 127.0.0.1:33271 2024-12-17 09:09:21,421 INFO hdfs.MiniDFSCluster: Cluster is active 2024-12-17 09:09:21,424 INFO mapreduce.MiniHadoopClusterManager: Started MiniDFSCluster -- namenode on port 12222 2024-12-17 09:09:21,430 INFO BlockStateChange: BLOCK* processReport 0x3ebf95827e3ada7e: Processing first storage report for DS-ff9e683c-0e29-43e9-8929-08b6041aaf16 from datanode a208327e-a2f5-4ce9-88fe-d17d4b676b80 2024-12-17 09:09:21,432 INFO BlockStateChange: BLOCK* processReport 0x3ebf95827e3ada7e: from storage DS-ff9e683c-0e29-43e9-8929-08b6041aaf16 node DatanodeRegistration(127.0.0.1:33271, datanodeUuid=a208327e-a2f5-4ce9-88fe-d17d4b676b80, infoPort=41033, infoSecurePort=0, ipcPort=43389, storageInfo=lv=-57;cid=testClusterID;nsid=1931591070;c=1734390559495), blocks: 0, hasStaleStorage: true, processing time: 2 msecs, invalidatedBlocks: 0 2024-12-17 09:09:21,432 INFO BlockStateChange: BLOCK* processReport 0x3ebf95827e3ada7e: Processing first storage report for DS-cb7c3e35-f4f6-4e0e-80d1-8cca01bf4884 from datanode a208327e-a2f5-4ce9-88fe-d17d4b676b80 2024-12-17 09:09:21,432 INFO BlockStateChange: BLOCK* processReport 0x3ebf95827e3ada7e: from storage DS-cb7c3e35-f4f6-4e0e-80d1-8cca01bf4884 node DatanodeRegistration(127.0.0.1:33271, datanodeUuid=a208327e-a2f5-4ce9-88fe-d17d4b676b80, infoPort=41033, infoSecurePort=0, ipcPort=43389, storageInfo=lv=-57;cid=testClusterID;nsid=1931591070;c=1734390559495), blocks: 0, hasStaleStorage: false, processing time: 1 msecs, invalidatedBlocks: 0 2024-12-17 09:09:21,446 INFO datanode.DataNode: Successfully sent block report 0x3ebf95827e3ada7e to namenode: localhost/127.0.0.1:12222, containing 2 storage report(s), of which we sent 2. The reports had 0 total blocks and used 1 RPC(s). This took 3 msecs to generate and 27 msecs for RPC and NN processing. Got back one command: FinalizeCommand/5. 2024-12-17 09:09:21,447 INFO datanode.DataNode: Got finalize command for block pool BP-1007324782-172.17.0.2-1734390559495 2024-12-17 09:10:40,892 INFO namenode.FSEditLog: Number of transactions: 2 Total time for transactions(ms): 13 Number of transactions batched in Syncs: 0 Number of syncs: 2 SyncTimes(ms): 1 0 2024-12-17 09:10:40,908 INFO hdfs.StateChange: BLOCK* allocate blk_1073741825_1001, replicas=127.0.0.1:33271 for /02845_table_function_hdfs_filter_by_virtual_columns_test_dmncmxow.data1.tsv 2024-12-17 09:10:40,959 INFO datanode.DataNode: Receiving BP-1007324782-172.17.0.2-1734390559495:blk_1073741825_1001 src: /127.0.0.1:44270 dest: /127.0.0.1:33271 2024-12-17 09:10:40,986 INFO DataNode.clienttrace: src: /127.0.0.1:44270, dest: /127.0.0.1:33271, bytes: 2, op: HDFS_WRITE, cliID: libhdfs3_client_rand_0.610816_count_2_pid_679_tid_139983833912896, offset: 0, srvID: a208327e-a2f5-4ce9-88fe-d17d4b676b80, blockid: BP-1007324782-172.17.0.2-1734390559495:blk_1073741825_1001, duration(ns): 8148625 2024-12-17 09:10:40,987 INFO datanode.DataNode: PacketResponder: BP-1007324782-172.17.0.2-1734390559495:blk_1073741825_1001, type=LAST_IN_PIPELINE terminating 2024-12-17 09:10:40,988 INFO hdfs.StateChange: BLOCK* fsync: /02845_table_function_hdfs_filter_by_virtual_columns_test_dmncmxow.data1.tsv for libhdfs3_client_rand_0.610816_count_2_pid_679_tid_139983833912896 2024-12-17 09:10:40,993 INFO namenode.FSNamesystem: BLOCK* blk_1073741825_1001 is COMMITTED but not COMPLETE(numNodes= 0 < minimum = 1) in file /02845_table_function_hdfs_filter_by_virtual_columns_test_dmncmxow.data1.tsv 2024-12-17 09:10:41,397 INFO hdfs.StateChange: DIR* completeFile: /02845_table_function_hdfs_filter_by_virtual_columns_test_dmncmxow.data1.tsv is closed by libhdfs3_client_rand_0.610816_count_2_pid_679_tid_139983833912896 2024-12-17 09:10:41,804 INFO hdfs.StateChange: BLOCK* allocate blk_1073741826_1002, replicas=127.0.0.1:33271 for /02845_table_function_hdfs_filter_by_virtual_columns_test_dmncmxow.data2.tsv 2024-12-17 09:10:41,806 INFO datanode.DataNode: Receiving BP-1007324782-172.17.0.2-1734390559495:blk_1073741826_1002 src: /127.0.0.1:44286 dest: /127.0.0.1:33271 2024-12-17 09:10:41,810 INFO DataNode.clienttrace: src: /127.0.0.1:44286, dest: /127.0.0.1:33271, bytes: 2, op: HDFS_WRITE, cliID: libhdfs3_client_rand_0.481618_count_4_pid_679_tid_139980443539008, offset: 0, srvID: a208327e-a2f5-4ce9-88fe-d17d4b676b80, blockid: BP-1007324782-172.17.0.2-1734390559495:blk_1073741826_1002, duration(ns): 2702670 2024-12-17 09:10:41,811 INFO datanode.DataNode: PacketResponder: BP-1007324782-172.17.0.2-1734390559495:blk_1073741826_1002, type=LAST_IN_PIPELINE terminating 2024-12-17 09:10:41,811 INFO hdfs.StateChange: BLOCK* fsync: /02845_table_function_hdfs_filter_by_virtual_columns_test_dmncmxow.data2.tsv for libhdfs3_client_rand_0.481618_count_4_pid_679_tid_139980443539008 2024-12-17 09:10:41,813 INFO hdfs.StateChange: DIR* completeFile: /02845_table_function_hdfs_filter_by_virtual_columns_test_dmncmxow.data2.tsv is closed by libhdfs3_client_rand_0.481618_count_4_pid_679_tid_139980443539008 2024-12-17 09:10:42,224 INFO hdfs.StateChange: BLOCK* allocate blk_1073741827_1003, replicas=127.0.0.1:33271 for /02845_table_function_hdfs_filter_by_virtual_columns_test_dmncmxow.data3.tsv 2024-12-17 09:10:42,226 INFO datanode.DataNode: Receiving BP-1007324782-172.17.0.2-1734390559495:blk_1073741827_1003 src: /127.0.0.1:44296 dest: /127.0.0.1:33271 2024-12-17 09:10:42,230 INFO DataNode.clienttrace: src: /127.0.0.1:44296, dest: /127.0.0.1:33271, bytes: 2, op: HDFS_WRITE, cliID: libhdfs3_client_rand_0.352420_count_6_pid_679_tid_139980433258048, offset: 0, srvID: a208327e-a2f5-4ce9-88fe-d17d4b676b80, blockid: BP-1007324782-172.17.0.2-1734390559495:blk_1073741827_1003, duration(ns): 1883404 2024-12-17 09:10:42,230 INFO datanode.DataNode: PacketResponder: BP-1007324782-172.17.0.2-1734390559495:blk_1073741827_1003, type=LAST_IN_PIPELINE terminating 2024-12-17 09:10:42,230 INFO hdfs.StateChange: BLOCK* fsync: /02845_table_function_hdfs_filter_by_virtual_columns_test_dmncmxow.data3.tsv for libhdfs3_client_rand_0.352420_count_6_pid_679_tid_139980433258048 2024-12-17 09:10:42,232 INFO hdfs.StateChange: DIR* completeFile: /02845_table_function_hdfs_filter_by_virtual_columns_test_dmncmxow.data3.tsv is closed by libhdfs3_client_rand_0.352420_count_6_pid_679_tid_139980433258048